15 research outputs found

    Fast fallback watermark detection using perceptual hashes

    Get PDF
    Forensic watermarking is often used to enable the tracing of digital pirates that leak copyright-protected videos. However, existing watermarking methods have a limited robustness and may be vulnerable to targeted attacks. Our previous work proposed a fallback detection method that uses secondary watermarks rather than the primary watermarks embedded by existing methods. However, the previously proposed fallback method is slow and requires access to all watermarked videos. This paper proposes to make the fallback watermark detection method faster using perceptual hashes instead of uncompressed secondary watermark signals. These perceptual hashes can be calculated prior to detection, such that the actual detection process is sped up with a factor of approximately 26,000 to 92,000. In this way, the proposed method tackles the main criticism about practical usability of the slow fallback method. The fast detection comes at the cost of a modest decrease in robustness, although the fast fallback detection method can still outperform the existing primary watermark method. In conclusion, the proposed method enables fast and more robust detection of watermarks that were embedded by existing watermarking methods

    On the impact of video stalling and video quality in the case of camera switching during adaptive streaming of sports content

    Get PDF
    The widespread usage of second screens, in combination with mobile video streaming technologies like HTTP Adaptive Streaming (HAS), enable new means for taking end-users' Quality of Experience (QoE) to the next level. For sports events, these technological evolutions can, for example, enhance the overall engagement of remote fans or give them more control over the content. In this paper, we consider the case of adaptively streaming multi-camera sports content to tablet devices, enabling the end-user to dynamically switch cameras. Our goal is to subjectively evaluate the trade-off between video stalling duration (as a result of requesting another camera feed) and initial video quality of the new feed. Our results show that short video stallings do not significantly influence overall quality ratings, that quality perception is highly influenced by the video quality at the moment of camera switching and that large quality fluctuations should be avoided

    Fast encoding for personalized views extracted from beyond high definition content

    Get PDF
    Broadcast providers are looking for new opportunities to increase user experience and user interaction on their content. Their main goal is to attract and preserve viewer attention to create a big and stable audience. This could be achieved with a second screen application that lets the users select their own viewpoint in an extremely high resolution video to direct their own first screen. By allowing the users to create their own personalized video stream, they become involved with the content creation itself. However, encoding a personalized view for each user is computationally complex. This paper describes a machine learning approach to speed up the encoding of each personal view. Simulation results of zoom, pan and tilt scenarios show bit rate increases between 2% and 9% for complexity reductions between 69% and 79% compared to full encoding

    Perceptual quality of 4K-resolution video content compared to HD

    Get PDF
    With the introduction of 4K UHD video and display resolution, questions arise on the perceptual differences between 4K UHD and upsampled HD video content. In this paper, a striped pair comparison has been performed on a diverse set of 4K UHD video sources. The goal was to subjectively assess the perceived sharpness of 4K UHD and downscaled/upscaled HD video. A striped pair comparison has been applied in order to make the test as straightforward as possible for a non-expert participant population. Under these conditions and over this set of sequences, on average, on 54.8% of the sequences (17 out of 31), 4K UHD resolution content could be identified as being sharper compared to its HD down and upsampled alternative. The probabilities in which 4K UHD could be differentiated from downscaled/upscaled HD range from 83.3% for the easiest to assess sequence down to 39.7% for the most difficult sequence. Although significance tests demonstrate there is a positive sharpness difference from camera quality 4K UHD content compared to the HD downscaled/upscaled variations, it is very content dependent and all circumstances have been chosen in favor of the 4K UHD representation. The results of this test can contribute to the research process of developing metrics indicating visibility of high resolution features within specific content

    The crowd as a cameraman : on-stage display of crowdsourced mobile video at large-scale events

    Get PDF
    Recording videos with smartphones at large-scale events such as concerts and festivals is very common nowadays. These videos register the atmosphere of the event as it is experienced by the crowd and offer a perspective that is hard to capture by the professional cameras installed throughout the venue. In this article, we present a framework to collect videos from smartphones in the public and blend these into a mosaic that can be readily mixed with professional camera footage and shown on displays during the event. The video upload is prioritized by matching requests of the event director with video metadata, while taking into account the available wireless network capacity. The proposed framework's main novelty is its scalability, supporting the real-time transmission, processing and display of videos recorded by hundreds of simultaneous users in ultra-dense Wi-Fi environments, as well as its proven integration in commercial production environments. The framework has been extensively validated in a controlled lab setting with up to 1 000 clients as well as in a field trial where 1 183 videos were collected from 135 participants recruited from an audience of 8 050 people. 90 % of those videos were uploaded within 6.8 minutes

    Fast simultaneous video encoder for adaptive streaming

    No full text
    Content providers create different versions of a video to accommodate different end-user devices and network conditions. However, each of these versions requires a resource intensive encoding process. To reduce the computational complexity of the encodings, this paper proposes a fast simultaneous encoder. This encoder takes a single video as input and creates a number of bit streams encoded with different parameters. Only one version of the video is created with a full encode, whereas encoding of the other versions is accelerated by exploiting the correlation with the fully encoded version using machine learning techniques. In a practical scenario, the fast simultaneous encoder achieves a complexity reduction of 67.3% with a bit rate increase of 5.2% compared to performing a full encode of each version

    Linking the cognitive load induced by route instruction types and building configuration during indoor route guidance, a usability study in VR

    No full text
    Every route instruction type (e.g. map, symbol, photo) induces a specific cognitive load. However, when these types are used at different decision points in a building, the building configuration of these points also influences the induced cognitive load. Therefore, the process of route guidance results in an interaction between the instruction type and the decision point, which determines the induced cognitive load. One way of reducing cognitive load during route guidance is by using adaptive systems that show specific route instruction types at specific decision points. Therefore, in this VR experiment, the usability of such an adaptive indoor route guidance system is tested by tracking the wayfinding and gaze behavior of the users. First, the difference in wayfinding and gaze behavior between all route instruction types is compared. Next, the building configuration at the decision points is quantified through the architectural theory of space syntax, and the correlation with the wayfinding and gaze behavior is determined. Our findings indicate that adapting the route instruction type does make a difference for the user
    corecore